There are multiple scales of abstraction from which we can describe the same image, depending on whether we are focusing on fine-grained details or a more global attribute of the image. In brain mapping, learning to automatically parse images to build representations of both small-scale features (e.g., the presence of cells or blood vessels) and global properties of an image (e.g., which brain region the image comes from) is a crucial and open challenge. However, most existing datasets and benchmarks for neuroanatomy consider only a single downstream task at a time. To bridge this gap, we introduce a new dataset, annotations, and multiple downstream tasks that provide diverse ways to readout information about brain structure and architecture from the same image. Our multi-task neuroimaging benchmark (MTNeuro) is built on volumetric, micrometer-resolution X-ray microtomography images spanning a large thalamocortical section of mouse brain, encompassing multiple cortical and subcortical regions. We generated a number of different prediction challenges and evaluated several supervised and self-supervised models for brain-region prediction and pixel-level semantic segmentation of microstructures. Our experiments not only highlight the rich heterogeneity of this dataset, but also provide insights into how self-supervised approaches can be used to learn representations that capture multiple attributes of a single image and perform well on a variety of downstream tasks. Datasets, code, and pre-trained baseline models are provided at: https://mtneuro.github.io/ .
translated by 谷歌翻译
Artificial Intelligence (AI) has become commonplace to solve routine everyday tasks. Because of the exponential growth in medical imaging data volume and complexity, the workload on radiologists is steadily increasing. We project that the gap between the number of imaging exams and the number of expert radiologist readers required to cover this increase will continue to expand, consequently introducing a demand for AI-based tools that improve the efficiency with which radiologists can comfortably interpret these exams. AI has been shown to improve efficiency in medical-image generation, processing, and interpretation, and a variety of such AI models have been developed across research labs worldwide. However, very few of these, if any, find their way into routine clinical use, a discrepancy that reflects the divide between AI research and successful AI translation. To address the barrier to clinical deployment, we have formed MONAI Consortium, an open-source community which is building standards for AI deployment in healthcare institutions, and developing tools and infrastructure to facilitate their implementation. This report represents several years of weekly discussions and hands-on problem solving experience by groups of industry experts and clinicians in the MONAI Consortium. We identify barriers between AI-model development in research labs and subsequent clinical deployment and propose solutions. Our report provides guidance on processes which take an imaging AI model from development to clinical implementation in a healthcare institution. We discuss various AI integration points in a clinical Radiology workflow. We also present a taxonomy of Radiology AI use-cases. Through this report, we intend to educate the stakeholders in healthcare and AI (AI researchers, radiologists, imaging informaticists, and regulators) about cross-disciplinary challenges and possible solutions.
translated by 谷歌翻译
机器学习(ML)的法律文献倾向于关注危害,因此倾向于推理个人模型结果和摘要错误率。这种关注模型级别的结果和错误掩盖了ML的重要方面,这些方面源于其固有的非确定性。我们表明,从关于ML输出作为可能结果的概率分布的推理的角度来看,非确定性的影响及其对法律的影响,对法律的影响变得更加清晰。这种分布观点通过强调ML的可能结果来解释非确定性。重要的是,这种推理并不是当前法律推理的独家性。它补充了(实际上可以加强)关于个人自动决策的个人,具体结果的分析。通过阐明非确定性的重要作用,我们证明了ML代码不在网络法线将“代码为法律视为法律”的框架之外,因为该框架假定代码是确定性的。最后,我们简要讨论了ML可以采取什么措施来限制非决定性造成危害的影响,并阐明法律必须在何处弥合其当前个人结果重点与分配方法之间的差距我们推荐。
translated by 谷歌翻译
肌肉驱动控制是跨越不同领域的兴趣的研究课题,特别是生物力学,机器人和图形。这种类型的控制尤其具有挑战性,因为模型通常是过度的,并且动态被延迟和非线性。然而,这是一个非常良好的测试和调整的致动模型,该模型经历了数百万年的演变,并且涉及有趣的性质利用肌肉肌腱单元的被动力和有效的能量存储和释放。为了促进肌肉致动模拟研究,我们基于Mujoco模拟器释放鸵鸟的3D肌肉骨骼模拟。 Ostriches是地球上最快的搭配之一,因此是研究肌肉驱动的双模运动的优秀模型。该模型基于CT扫描和解剖,用于收集诸如插入位点,长度和钢圈角度的实际肌肉数据。除此之外,我们还提供一组加强学习任务,包括参考运动跟踪和颈部的达到任务。参考运动数据基于我们预处理和适应我们模型的各种行为的运动捕获剪辑。本文介绍了如何使用任务构建和迭代地改进模型。通过将它们与从机车鸟类的实验收集的电拍摄数据进行比较来评估肌肉致动模式的准确性。我们认为,这项工作可以是生物力学,强化学习,图形和机器人社区之间的有用桥梁,通过提供快速且易于使用的模拟。
translated by 谷歌翻译
人工智能将人工智能融入临床工作流程需要可靠且强大的模型。鲁棒性的主要特征是可重复性。在不评估模型重复性的情况下,给予分类性能很多,导致在实践中不可用的模型开发。在这项工作中,我们评估了在同一访问期间获得的同一患者的四种模型类型的可重复性。我们研究了三个医学图像分析任务的二进制,多级,序数和回归模型的性能:宫颈癌筛查,乳房密度估计和早产分类视网膜病变。此外,我们评估采样蒙特卡罗辍学预测在分类性能和可重复性上的测试时间的影响。利用Monte Carlo预测,为二元,多级和序数模型的所有任务的重复性显着提高,导致平均减少95%协议限额17%的分数。
translated by 谷歌翻译
尽管对临床机器学习研究有强烈的关注和相当大的投资,但在现实世界的临床环境中,在大规模的应用中已经部署了相对较少的应用。虽然研究在推进最先进的情况下很重要,但翻译同样重要的是,使这些技术和技术能够最终影响医疗保健。我们认为对几个考虑缺乏升值是在期望和现实之间这种差异的主要原因。为了更好地描述研究人员和从业者之间的整体视角,我们调查了几个从业人员在开发CML中进行临床部署的商业经验。使用这些洞察力,我们确定了几个主要类别的挑战,以便更好地设计和开发临床机学习应用。
translated by 谷歌翻译
在地质不确定性下,快速同化监测数据以更新压力累积和压力累积和二氧化碳(CO2)羽流迁移的预测是地质碳储存中的一个具有挑战性的问题。具有高维参数空间的数据同化的高计算成本阻碍了商业规模库管理的快速决策。我们建议利用具有深度学习技术的多孔介质流动行为的物理理解,以开发快速历史匹配 - 水库响应预测工作流程。应用集合更顺畅的多数据同化框架,工作流程更新地质特性,并通过通过地震反转解释的压力历史和二氧化碳羽毛的量化不确定性来预测水库性能。由于这种工作流程中最具计算昂贵的组件是储层模拟,我们开发了代理模型,以在多孔注射下预测动态压力和CO2羽流量。代理模型采用深度卷积神经网络,具体地,宽的剩余网络和残留的U-Net。该工作流程针对代表碎屑货架沉积环境的扁平三维储层模型验证。智能处理应用于真正的3D储层模型中数量与单层储层模型之间的桥梁。工作流程可以在主流个人工作站上不到一小时内完成历史匹配和储库预测,在不到一小时内。
translated by 谷歌翻译
我们探索使用大型预用语言模型作为少量语义解析器。语义解析中的目标是给定自然语言输入的结构化含义表示。但是,培训语言模型以生成自然语言。为了弥合差距,我们使用语言模型来解释进入一个类似于英语的受控的子宫内的输入,可以自动映射到目标含义表示表示。我们的结果表明,只有少量的数据和较少的代码转换为类似英语的代表,我们为快速启动语义解析器的蓝图导致了对多个社区任务的令人惊讶的有效性能,大大超过基线方法也在相同的限制上培训数据。
translated by 谷歌翻译
Remote sensing imagery provides comprehensive views of the Earth, where different sensors collect complementary data at different spatial scales. Large, pretrained models are commonly finetuned with imagery that is heavily augmented to mimic different conditions and scales, with the resulting models used for various tasks with imagery from a range of spatial scales. Such models overlook scale-specific information in the data. In this paper, we present Scale-MAE, a pretraining method that explicitly learns relationships between data at different, known scales throughout the pretraining process. Scale-MAE pretrains a network by masking an input image at a known input scale, where the area of the Earth covered by the image determines the scale of the ViT positional encoding, not the image resolution. Scale-MAE encodes the masked image with a standard ViT backbone, and then decodes the masked image through a bandpass filter to reconstruct low/high frequency images at lower/higher scales. We find that tasking the network with reconstructing both low/high frequency images leads to robust multiscale representations for remote sensing imagery. Scale-MAE achieves an average of a $5.0\%$ non-parametric kNN classification improvement across eight remote sensing datasets compared to current state-of-the-art and obtains a $0.9$ mIoU to $3.8$ mIoU improvement on the SpaceNet building segmentation transfer task for a range of evaluation scales.
translated by 谷歌翻译
With an ever-growing number of parameters defining increasingly complex networks, Deep Learning has led to several breakthroughs surpassing human performance. As a result, data movement for these millions of model parameters causes a growing imbalance known as the memory wall. Neuromorphic computing is an emerging paradigm that confronts this imbalance by performing computations directly in analog memories. On the software side, the sequential Backpropagation algorithm prevents efficient parallelization and thus fast convergence. A novel method, Direct Feedback Alignment, resolves inherent layer dependencies by directly passing the error from the output to each layer. At the intersection of hardware/software co-design, there is a demand for developing algorithms that are tolerable to hardware nonidealities. Therefore, this work explores the interrelationship of implementing bio-plausible learning in-situ on neuromorphic hardware, emphasizing energy, area, and latency constraints. Using the benchmarking framework DNN+NeuroSim, we investigate the impact of hardware nonidealities and quantization on algorithm performance, as well as how network topologies and algorithm-level design choices can scale latency, energy and area consumption of a chip. To the best of our knowledge, this work is the first to compare the impact of different learning algorithms on Compute-In-Memory-based hardware and vice versa. The best results achieved for accuracy remain Backpropagation-based, notably when facing hardware imperfections. Direct Feedback Alignment, on the other hand, allows for significant speedup due to parallelization, reducing training time by a factor approaching N for N-layered networks.
translated by 谷歌翻译